Multi-Instance Dimensionality Reduction

نویسندگان

  • Yu-Yin Sun
  • Michael K. Ng
  • Zhi-Hua Zhou
چکیده

Multi-instance learning deals with problems that treat bags of instances as training examples. In single-instance learning problems, dimensionality reduction is an essential step for high-dimensional data analysis and has been studied for years. The curse of dimensionality also exists in multiinstance learning tasks, yet this difficult task has not been studied before. Direct application of existing single-instance dimensionality reduction objectives to multi-instance learning tasks may not work well since it ignores the characteristic of multi-instance learning that the labels of bags are known while the labels of instances are unknown. In this paper, we propose an effective model and develop an efficient algorithm to solve the multi-instance dimensionality reduction problem. We formulate the objective as an optimization problem by considering orthonormality and sparsity constraints in the projection matrix for dimensionality reduction, and then solve it by the gradient descent along the tangent space of the orthonormal matrices. We also propose an approximation for improving the efficiency. Experimental results validate the effectiveness of the proposed method. Introduction In single-instance scenario we are given a training set containing N instances with their labels. In multi-instance learning (Dietterich, Lathrop, and Lozano-Perez 1997) the training examples are N bags each containing many instances. The labels of training bags are known yet the labels of the training instances are unknown. According to the standard multi-instance learning assumption, a positive bag contains at least one positive instance, while all the instances in negative bags are negative. Multi-instance learning has been found useful in modeling many real world applications such as drug activity prediction (Dietterich, Lathrop, and Lozano-Perez 1997), image retrieval (Andrews, Tsochantaridis, and Hofmann 2003), text categorization (Andrews, Tsochantaridis, and Hofmann 2003), face detection (Viola, Platt, and Zhang 2006), computer-aidedmedical diagnosis (Fung et al. 2007), ∗Supported by the National Fundamental Research Program of China (2010CB327903), the National Science Foundation of China (60635030, 60721002) and Jiangsu Science Foundation (BK2008018). Copyright c © 2010, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. etc. Many of these tasks involve high-dimensional data and thus encounter the curse of dimensionality. In single-instance scenario the curse of dimensionality has attracted much attention. There are two major paradigms, i.e., feature selection and dimensionality reduction. Feature selection tries to select a subset of the original features according to some measurements such as the mutual information or distance-based measures. Raykar et al. (2008) have studied multi-instance feature selection using Bayesian method which automatically considers the feature relevance. In most cases, searching an optimal feature subset is hard and heuristic methods are often used. Dimensionality reduction, which tries to extract a small number of new features by projecting the original features into a new space, is generally with better theoretical foundation. Existing dimensionality reduction techniques can be roughly divided into two categories, that is, unsupervised approaches such as PCA (principal component analysis) (Jolliffe 2002), and supervised approaches such as LDA (linear discriminant analysis) (Fukunaga 1990). To the best of our knowledge, multiinstance dimensionality reduction has not been studied before. It is noteworthy that multi-instance dimensionality reduction is even harder than single-instance dimensionality reduction since the input space of multi-instance learning task is ambiguous. In this paper, we propose the MIDR (Multi-Instance Dimensionality Reduction) approach based on a specifically designed dimensionality reduction objective for multiinstance learning. We formulate the objective as an optimization problem by considering orthonormality and sparsity constraints in the projection matrix for dimensionality reduction, and then solve it by gradient descent along the tangent space of the orthonormal matrices. We also propose an approximation to improve the efficiency. Experimental results validate the effectiveness of the proposed method. The rest of this paper is organized as follows. We start by a brief review of related work. Then, we propose MIDR and report our experiments, which is followed by the conclusion.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Non-I.I.D. Multi-Instance Dimensionality Reduction by Learning a Maximum Bag Margin Subspace

Multi-instance learning, as other machine learning tasks, also suffers from the curse of dimensionality. Although dimensionality reduction methods have been investigated for many years, multi-instance dimensionality reduction methods remain untouched. On the other hand, most algorithms in multi-instance framework treat instances in each bag as independently and identically distributed (i.i.d.) ...

متن کامل

Is Multinomial PCA Multi-faceted Clustering or Dimensionality Reduction?

Discrete analogues to Principal Components Analysis (PCA) are intended to handle discrete or positive-only data, for instance sets of documents. The class of methods is appropriately called multinomial PCA because it replaces the Gaussian in the probabilistic formulation of PCA with a multinomial. Experiments to date, however, have been on small data sets, for instance, from early information r...

متن کامل

Dimensionality Reduction and Improving the Performance of Automatic Modulation Classification using Genetic Programming (RESEARCH NOTE)

This paper shows how we can make advantage of using genetic programming in selection of suitable features for automatic modulation recognition. Automatic modulation recognition is one of the essential components of modern receivers. In this regard, selection of suitable features may significantly affect the performance of the process. Simulations were conducted with 5db and 10db SNRs. Test and ...

متن کامل

READER: Robust Semi-Supervised Multi-Label Dimension Reduction

Multi-label classification is an appealing and challenging supervised learning problem, where multiple labels, rather than a single label, are associated with an unseen test instance. To remove possible noises in labels and features of high-dimensionality, multi-label dimension reduction has attracted more and more attentions in recent years. The existing methods usually suffer from several pro...

متن کامل

Exploiting Multi-Label Information for Noise Resilient Feature Selection

In conventional supervised learning paradigm, each data instance is associated with one single class label. Multi-label learning differs in the way that data instances may belong to multiple concepts simultaneously, which naturally appear in a variety of high impact domains, ranging from bioinformatics, information retrieval to multimedia analysis. It targets to leverage the multiple label info...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2010